Constrained Risk-Averse Markov Decision Processes

نویسندگان

چکیده

We consider the problem of designing policies for Markov decision processes (MDPs) with dynamic coherent risk objectives and constraints. begin by formulating in a Lagrangian framework. Under assumption that constraints can be represented transition mapping, we propose an optimization-based method to synthesize Markovian lower-bound constrained risk-averse problem. demonstrate formulated optimization problems are form difference convex programs (DCPs) solved disciplined convex-concave programming (DCCP) show these results generalize linear MDPs total discounted expected costs Finally, illustrate effectiveness proposed numerical experiments on rover navigation involving conditional-value-at-risk (CVaR) entropic-value-at-risk (EVaR) measures.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Risk-averse dynamic programming for Markov decision processes

We introduce the concept of a Markov risk measure and we use it to formulate risk-averse control problems for two Markov decision models: a finite horizon model and a discounted infinite horizon model. For both models we derive risk-averse dynamic programming equations and a value iteration method. For the infinite horizon problem we also develop a risk-averse policy iteration method and we pro...

متن کامل

Constrained Markov Decision Processes

2 i To Tania and Einat ii Preface In many situations in the optimization of dynamic systems, a single utility for the optimizer might not suuce to describe the real objectives involved in the sequential decision making. A natural approach for handling such cases is that of optimization of one objective with constraints on other ones. This allows in particular to understand the tradeoo between t...

متن کامل

Hard Constrained Semi-Markov Decision Processes

In multiple criteria Markov Decision Processes (MDP) where multiple costs are incurred at every decision point, current methods solve them by minimising the expected primary cost criterion while constraining the expectations of other cost criteria to some critical values. However, systems are often faced with hard constraints where the cost criteria should never exceed some critical values at a...

متن کامل

Relaxation for Constrained Decentralized Markov Decision Processes

This paper studies a class of decentralized multi-agent stochastic optimization problems. In these problems, each agent has only a partial view of the world state, and a partial control of the actions but must cooperatively maximize the long-term system reward. The state that an agent observe consists of two parts a common public component and an agentspecific private component. Importantly, ta...

متن کامل

Dynamic programming in constrained Markov decision processes

We consider a discounted Markov Decision Process (MDP) supplemented with the requirement that another discounted loss must not exceed a specified value, almost surely. We show that the problem can be reformulated as a standard MDP and solved using the Dynamic Programming approach. An example on a controlled queue is presented. In the last section, we briefly reinforce the connection of the Dyna...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2021

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v35i13.17393